18 research outputs found
Explaining Trained Neural Networks with Semantic Web Technologies: First Steps
The ever increasing prevalence of publicly available structured data on the
World Wide Web enables new applications in a variety of domains. In this paper,
we provide a conceptual approach that leverages such data in order to explain
the input-output behavior of trained artificial neural networks. We apply
existing Semantic Web technologies in order to provide an experimental proof of
concept
Explaining Deep Learning Hidden Neuron Activations using Concept Induction
One of the current key challenges in Explainable AI is in correctly
interpreting activations of hidden neurons. It seems evident that accurate
interpretations thereof would provide insights into the question what a deep
learning system has internally \emph{detected} as relevant on the input, thus
lifting some of the black box character of deep learning systems.
The state of the art on this front indicates that hidden node activations
appear to be interpretable in a way that makes sense to humans, at least in
some cases. Yet, systematic automated methods that would be able to first
hypothesize an interpretation of hidden neuron activations, and then verify it,
are mostly missing.
In this paper, we provide such a method and demonstrate that it provides
meaningful interpretations. It is based on using large-scale background
knowledge -- a class hierarchy of approx. 2 million classes curated from the
Wikipedia Concept Hierarchy -- together with a symbolic reasoning approach
called \emph{concept induction} based on description logics that was originally
developed for applications in the Semantic Web field.
Our results show that we can automatically attach meaningful labels from the
background knowledge to individual neurons in the dense layer of a
Convolutional Neural Network through a hypothesis and verification process.Comment: Submitted to IJCAI-2
Neuro-Symbolic Deductive Reasoning for Cross-Knowledge Graph Entailment
A significant and recent development in neural-symbolic learning are deep neural networks that can reason over symbolic knowledge graphs (KGs). A particular task of interest is KG entailment, which is to infer the set of all facts that are a logical consequence of current and potential facts of a KG. Initial neural-symbolic systems that can deduce the entailment of a KG have been presented, but they are limited: current systems learn fact relations and entailment patterns specific to a particular KG and hence do not truly generalize, and must be retrained for each KG they are tasked with entailing. We propose a neural-symbolic system to address this limitation in this paper. It is designed as a differentiable end-to-end deep memory network that learns over abstract, generic symbols to discover entailment patterns common to any reasoning task. A key component of the system is a simple but highly effective normalization process for continuous representation learning of KG entities within memory networks. Our results show how the model, trained over a set of KGs, can effectively entail facts from KGs excluded from the training, even when the vocabulary or the domain of test KGs is completely different from the training KGs
Improved Javanese script recognition using custom model of convolution neural network
Handwriting recognition in Javanese script is not widely developed with deep learning (DL). Previous DL and machine learning (ML) research is generally limited to basic characters (Carakan) only. This study proposes a deep learning model using a custom-built convolutional neural network to improve recognition accuracy performance and reduce computational costs. The main features of handwritten objects are textures, edges, lines, and shapes, so convolution layers are not designed in large numbers. This research maximizes optimization of other layers such as pooling, activation function, fully connected layer, optimizer, and parameter settings such as dropout and learning rate. There are eleven main layers used in the proposed custom convolutional neural network (CNN) model, namely four convolution layers+activation function, four pooling layers, two fully connected layers, and a softmax classifier. Based on the test results on the Javanese script handwritten image dataset with 120 classes consisting of 20 basic character classes and 100 compound character classes, the resulting accuracy is 97.29%